30 research outputs found

    Learning Geometric Concepts with Nasty Noise

    Full text link
    We study the efficient learnability of geometric concept classes - specifically, low-degree polynomial threshold functions (PTFs) and intersections of halfspaces - when a fraction of the data is adversarially corrupted. We give the first polynomial-time PAC learning algorithms for these concept classes with dimension-independent error guarantees in the presence of nasty noise under the Gaussian distribution. In the nasty noise model, an omniscient adversary can arbitrarily corrupt a small fraction of both the unlabeled data points and their labels. This model generalizes well-studied noise models, including the malicious noise model and the agnostic (adversarial label noise) model. Prior to our work, the only concept class for which efficient malicious learning algorithms were known was the class of origin-centered halfspaces. Specifically, our robust learning algorithm for low-degree PTFs succeeds under a number of tame distributions -- including the Gaussian distribution and, more generally, any log-concave distribution with (approximately) known low-degree moments. For LTFs under the Gaussian distribution, we give a polynomial-time algorithm that achieves error O(ϵ)O(\epsilon), where ϵ\epsilon is the noise rate. At the core of our PAC learning results is an efficient algorithm to approximate the low-degree Chow-parameters of any bounded function in the presence of nasty noise. To achieve this, we employ an iterative spectral method for outlier detection and removal, inspired by recent work in robust unsupervised learning. Our aforementioned algorithm succeeds for a range of distributions satisfying mild concentration bounds and moment assumptions. The correctness of our robust learning algorithm for intersections of halfspaces makes essential use of a novel robust inverse independence lemma that may be of broader interest

    Optimal Testing of Discrete Distributions with High Probability

    Get PDF

    Optimal Testing of Discrete Distributions with High Probability

    Get PDF
    We study the problem of testing discrete distributions with a focus on the high probability regime. Specifically, given samples from one or more discrete distributions, a property P\mathcal{P}, and parameters 0<ϵ,δ<10< \epsilon, \delta <1, we want to distinguish {\em with probability at least 1δ1-\delta} whether these distributions satisfy P\mathcal{P} or are ϵ\epsilon-far from P\mathcal{P} in total variation distance. Most prior work in distribution testing studied the constant confidence case (corresponding to δ=Ω(1)\delta = \Omega(1)), and provided sample-optimal testers for a range of properties. While one can always boost the confidence probability of any such tester by black-box amplification, this generic boosting method typically leads to sub-optimal sample bounds. Here we study the following broad question: For a given property P\mathcal{P}, can we {\em characterize} the sample complexity of testing P\mathcal{P} as a function of all relevant problem parameters, including the error probability δ\delta? Prior to this work, uniformity testing was the only statistical task whose sample complexity had been characterized in this setting. As our main results, we provide the first algorithms for closeness and independence testing that are sample-optimal, within constant factors, as a function of all relevant parameters. We also show matching information-theoretic lower bounds on the sample complexity of these problems. Our techniques naturally extend to give optimal testers for related problems. To illustrate the generality of our methods, we give optimal algorithms for testing collections of distributions and testing closeness with unequal sized samples

    On the Complexity of Optimal Lottery Pricing and Randomized Mechanisms

    Get PDF

    The Fourier Transform of Poisson Multinomial Distributions and its Algorithmic Applications

    Get PDF
    An (n,k)(n, k)-Poisson Multinomial Distribution (PMD) is a random variable of the form X=i=1nXiX = \sum_{i=1}^n X_i, where the XiX_i's are independent random vectors supported on the set of standard basis vectors in Rk.\mathbb{R}^k. In this paper, we obtain a refined structural understanding of PMDs by analyzing their Fourier transform. As our core structural result, we prove that the Fourier transform of PMDs is {\em approximately sparse}, i.e., roughly speaking, its L1L_1-norm is small outside a small set. By building on this result, we obtain the following applications: {\bf Learning Theory.} We design the first computationally efficient learning algorithm for PMDs with respect to the total variation distance. Our algorithm learns an arbitrary (n,k)(n, k)-PMD within variation distance ϵ\epsilon using a near-optimal sample size of O~k(1/ϵ2),\widetilde{O}_k(1/\epsilon^2), and runs in time O~k(1/ϵ2)logn.\widetilde{O}_k(1/\epsilon^2) \cdot \log n. Previously, no algorithm with a poly(1/ϵ)\mathrm{poly}(1/\epsilon) runtime was known, even for k=3.k=3. {\bf Game Theory.} We give the first efficient polynomial-time approximation scheme (EPTAS) for computing Nash equilibria in anonymous games. For normalized anonymous games with nn players and kk strategies, our algorithm computes a well-supported ϵ\epsilon-Nash equilibrium in time nO(k3)(k/ϵ)O(k3log(k/ϵ)/loglog(k/ϵ))k1.n^{O(k^3)} \cdot (k/\epsilon)^{O(k^3\log(k/\epsilon)/\log\log(k/\epsilon))^{k-1}}. The best previous algorithm for this problem had running time n(f(k)/ϵ)k,n^{(f(k)/\epsilon)^k}, where f(k)=Ω(kk2)f(k) = \Omega(k^{k^2}), for any k>2.k>2. {\bf Statistics.} We prove a multivariate central limit theorem (CLT) that relates an arbitrary PMD to a discretized multivariate Gaussian with the same mean and covariance, in total variation distance. Our new CLT strengthens the CLT of Valiant and Valiant by completely removing the dependence on nn in the error bound.Comment: 68 pages, full version of STOC 2016 pape

    How Good is the Chord Algorithm?

    No full text
    corecore